chatgpt gpt-4
DeID-GPT: Zero-shot Medical Text De-Identification by GPT-4
Liu, Zhengliang, Huang, Yue, Yu, Xiaowei, Zhang, Lu, Wu, Zihao, Cao, Chao, Dai, Haixing, Zhao, Lin, Li, Yiwei, Shu, Peng, Zeng, Fang, Sun, Lichao, Liu, Wei, Shen, Dinggang, Li, Quanzheng, Liu, Tianming, Zhu, Dajiang, Li, Xiang
The digitization of healthcare has facilitated the sharing and re-using of medical data but has also raised concerns about confidentiality and privacy. HIPAA (Health Insurance Portability and Accountability Act) mandates removing re-identifying information before the dissemination of medical records. Thus, effective and efficient solutions for de-identifying medical data, especially those in free-text forms, are highly needed. While various computer-assisted de-identification methods, including both rule-based and learning-based, have been developed and used in prior practice, such solutions still lack generalizability or need to be fine-tuned according to different scenarios, significantly imposing restrictions in wider use. The advancement of large language models (LLM), such as ChatGPT and GPT-4, have shown great potential in processing text data in the medical domain with zero-shot in-context learning, especially in the task of privacy protection, as these models can identify confidential information by their powerful named entity recognition (NER) capability. In this work, we developed a novel GPT4-enabled de-identification framework (``DeID-GPT") to automatically identify and remove the identifying information. Compared to existing commonly used medical text data de-identification methods, our developed DeID-GPT showed the highest accuracy and remarkable reliability in masking private information from the unstructured medical text while preserving the original structure and meaning of the text. This study is one of the earliest to utilize ChatGPT and GPT-4 for medical text data processing and de-identification, which provides insights for further research and solution development on the use of LLMs such as ChatGPT/GPT-4 in healthcare. Codes and benchmarking data information are available at https://github.com/yhydhx/ChatGPT-API.
- North America > United States > Georgia > Clarke County > Athens (0.14)
- Asia > China > Shanghai > Shanghai (0.04)
- North America > United States > Texas > Tarrant County > Arlington (0.04)
- (6 more...)
- Research Report > Experimental Study (1.00)
- Overview (1.00)
- Research Report > New Finding (0.93)
The Potential and Pitfalls of using a Large Language Model such as ChatGPT or GPT-4 as a Clinical Assistant
Zhang, Jingqing, Sun, Kai, Jagadeesh, Akshay, Ghahfarokhi, Mahta, Gupta, Deepa, Gupta, Ashok, Gupta, Vibhor, Guo, Yike
Recent studies have demonstrated promising performance of ChatGPT and GPT-4 on several medical domain tasks. However, none have assessed its performance using a large-scale real-world electronic health record database, nor have evaluated its utility in providing clinical diagnostic assistance for patients across a full range of disease presentation. We performed two analyses using ChatGPT and GPT-4, one to identify patients with specific medical diagnoses using a real-world large electronic health record database and the other, in providing diagnostic assistance to healthcare workers in the prospective evaluation of hypothetical patients. Our results show that GPT-4 across disease classification tasks with chain of thought and few-shot prompting can achieve performance as high as 96% F1 scores. For patient assessment, GPT-4 can accurately diagnose three out of four times. However, there were mentions of factually incorrect statements, overlooking crucial medical findings, recommendations for unnecessary investigations and overtreatment. These issues coupled with privacy concerns, make these models currently inadequate for real world clinical use. However, limited data and time needed for prompt engineering in comparison to configuration of conventional machine learning workflows highlight their potential for scalability across healthcare applications.
- Europe > Austria > Vienna (0.14)
- Asia > China > Hong Kong (0.04)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- (2 more...)
Exploring the Trade-Offs: Unified Large Language Models vs Local Fine-Tuned Models for Highly-Specific Radiology NLI Task
Wu, Zihao, Zhang, Lu, Cao, Chao, Yu, Xiaowei, Dai, Haixing, Ma, Chong, Liu, Zhengliang, Zhao, Lin, Li, Gang, Liu, Wei, Li, Quanzheng, Shen, Dinggang, Li, Xiang, Zhu, Dajiang, Liu, Tianming
Recently, ChatGPT and GPT-4 have emerged and gained immense global attention due to their unparalleled performance in language processing. Despite demonstrating impressive capability in various open-domain tasks, their adequacy in highly specific fields like radiology remains untested. Radiology presents unique linguistic phenomena distinct from open-domain data due to its specificity and complexity. Assessing the performance of large language models (LLMs) in such specific domains is crucial not only for a thorough evaluation of their overall performance but also for providing valuable insights into future model design directions: whether model design should be generic or domain-specific. To this end, in this study, we evaluate the performance of ChatGPT/GPT-4 on a radiology NLI task and compare it to other models fine-tuned specifically on task-related data samples. We also conduct a comprehensive investigation on ChatGPT/GPT-4's reasoning ability by introducing varying levels of inference difficulty. Our results show that 1) GPT-4 outperforms ChatGPT in the radiology NLI task; 2) other specifically fine-tuned models require significant amounts of data samples to achieve comparable performance to ChatGPT/GPT-4. These findings demonstrate that constructing a generic model that is capable of solving various tasks across different domains is feasible.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > China > Shanghai > Shanghai (0.05)
- North America > United States > Texas (0.04)
- (8 more...)
- Health & Medicine > Nuclear Medicine (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
What Are the Data-Centric AI Concepts behind GPT Models?
Artificial Intelligence (AI) has made incredible strides in transforming the way we live, work, and interact with technology. Recently, that one area that has seen significant progress is the development of Large Language Models (LLMs), such as GPT-3, ChatGPT, and GPT-4. These models are capable of performing tasks such as language translation, text summarization, and question-answering with impressive accuracy. While it's difficult to ignore the increasing model size of LLMs, it's also important to recognize that their success is due largely to the large amount and high-quality data used to train them. In this article, we will present an overview of the recent advancements in LLMs from a data-centric AI perspective, drawing upon insights from our recent survey papers [1,2] with corresponding technical resources on GitHub.
Beyond GPT-4: The Importance of Building Custom ML Models
As the field of AI and machine learning continues to evolve, pre-trained language models like ChatGPT/GPT-4 have emerged as powerful tools for natural language processing tasks. In a recent tweet, Daniel Bourke posed the question, "Why bother building your own custom ML models when ChatGPT/GPT-4 will be better?" It's a valid question, given the impressive capabilities of pre-trained language models like ChatGPT/GPT-4. However, there are still several compelling reasons to build custom ML models, even in the face of such impressive technology. If you are interested in learning more about AI and machine learning, you may want to check out Daniel Bourke's Twitter and Medium accounts, as well as his YouTube channel.